Learning Model-Based Sparsity via Projected Gradient Descent
نویسندگان
چکیده
منابع مشابه
Spectral Compressed Sensing via Projected Gradient Descent
Let x ∈ C be a spectrally sparse signal consisting of r complex sinusoids with or without damping. We consider the spectral compressed sensing problem, which is about reconstructing x from its partial revealed entries. By utilizing the low rank structure of the Hankel matrix corresponding to x, we develop a computationally efficient algorithm for this problem. The algorithm starts from an initi...
متن کاملLearning ReLUs via Gradient Descent
In this paper we study the problem of learning Rectified Linear Units (ReLUs) which are functions of the form x ↦ max(0, ⟨w,x⟩) with w ∈ R denoting the weight vector. We study this problem in the high-dimensional regime where the number of observations are fewer than the dimension of the weight vector. We assume that the weight vector belongs to some closed set (convex or nonconvex) which captu...
متن کاملMaterial for “ Spectral Compressed Sensing via Projected Gradient Descent ”
We extend PGD and its recovery guarantee [1] from one-dimensional spectrally sparse signal recovery to the multi-dimensional case. Assume the underlying multi-dimensional spectrally sparse signal is of model order r and total dimension N . We show that O(r log(N)) measurements are sufficient for PGD to achieve successful recovery with high probability provided the underlying signal satisfies so...
متن کاملLearning via Gradient Descent in Sigma
Integrating a gradient-descent learning mechanism at the core of the graphical models upon which the Sigma cognitive architecture/system is built yields learning behaviors that span important forms of both procedural learning (e.g., action and reinforcement learning) and declarative learning (e.g., supervised and unsupervised concept formation), plus several additional forms of learning (e.g., ...
متن کاملVAE Learning via Stein Variational Gradient Descent
A new method for learning variational autoencoders (VAEs) is developed, based on Stein variational gradient descent. A key advantage of this approach is that one need not make parametric assumptions about the form of the encoder distribution. Performance is further enhanced by integrating the proposed encoder with importance sampling. Excellent performance is demonstrated across multiple unsupe...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Information Theory
سال: 2016
ISSN: 0018-9448,1557-9654
DOI: 10.1109/tit.2016.2515078